Self-organizing mixture models
نویسندگان
چکیده
We present an expectation-maximization (EM) algorithm that yields topology preserving maps of data based on probabilistic mixture models. Our approach is applicable to any mixture model for which we have a normal EM algorithm. Compared to other mixture model approaches to self-organizing maps, the function our algorithm maximizes has a clear interpretation: it sums data log-likelihood and a penalty term that enforces self-organization. Our approach allows principled handling of missing data and learning of mixtures of self-organizing maps. We present example applications illustrating our approach for continuous, discrete, and mixed discrete and continuous data.
منابع مشابه
Comparison of Self-Organizing Maps, Mixture, K-means and Hybrid Approaches to Risk Classification of Passive Railway Crossings
We create factor constructs for the physical characteristics of 864 passive railway crossings in South Australia. Crossings are then classified as dangerous or otherwise by means of self-organizing maps, K-means, mixture models and combinations of the latter. Results are compared using historical accident data. The self-organizing map with mixtures approach is found to be optimal in prediction ...
متن کاملSupporting Maritime Situation Awareness Using Self Organizing Maps and Gaussian Mixture Models
Maritime situation awareness is of importance in a lot of areas – e.g. detection of weapon smuggling in military peacekeeping operations, and harbor traffic control missions for the coast guard. In this paper, we have combined the use of Self Organizing Maps with Gaussian Mixture Models, in order to enable situation awareness by detecting deviations from normal behavior in an unsupervised way. ...
متن کاملEmbedded Classification Kernel Using Som Clustering and Mixture of Experts
-In this paper, we introduce a new classification kernel by embedding self organized map (SOM) clustering with mixture of radial basis function (RBF) networks. The model’s efficacy is demonstrated in solving a multi-class TIMIT speech recognition problem where the kernel is used to learn the multidimensional cepstral feature vectors to estimate their posterior class probabilities. The tests res...
متن کاملSelf-organization in mixture densities of HMM based speech recognition
In this paper experiments are presented to apply Self-Organizing Map (SOM) and Learning Vector Quantization (LVQ) for training mixture density hidden Markov models (HMMs) in automatic speech recognition. The decoding of spoken words into text is made using speaker dependent, but vocabulary and context independent phoneme HMMs. Each HMM has a set of states and the output density of each state is...
متن کاملSelf-Organization by Optimizing Free-Energy
We present a variational Expectation-Maximization algorithm to learn probabilistic mixture models. The algorithm is similar to Kohonen’s Self-Organizing Map algorithm and not limited to Gaussian mixtures. We maximize the variational free-energy that sums data loglikelihood and Kullback-Leibler divergence between a normalized neighborhood function and the posterior distribution on the components...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- Neurocomputing
دوره 63 شماره
صفحات -
تاریخ انتشار 2005